Multi-LLM router MCP server for Claude Code — smart complexity routing, Claude subscription monitoring, Codex integration, 20+ providers
Project description
LLM Router
One MCP server. Every AI model. Smart routing.
Route text, image, video, and audio tasks to 20+ AI providers — automatically picking the best model for the job based on your budget and active profile.
Quick Start • How It Works • Providers • Tools • Configuration • Provider Setup
The Problem
You use Claude Code. You also have GPT-4o, Gemini, Perplexity, DALL-E, Runway, ElevenLabs — but switching between them is manual, slow, and expensive.
LLM Router gives your AI assistant one unified interface to all of them — and automatically picks the right one based on what you're doing and what you can afford.
You: "Research the latest AI funding rounds"
Router: → Perplexity Sonar Pro (search-augmented, best for current facts)
You: "Generate a hero image for the landing page"
Router: → Flux Pro via fal.ai (best quality/cost for images)
You: "Write unit tests for the auth module"
Router: → Claude Sonnet (top coding model, within budget)
You: "Create a 5-second product demo clip"
Router: → Kling 2.0 via fal.ai (best value for short video)
How It Saves You Money
Not every task needs the same model. Without a router, everything goes to the same expensive model — like hiring a surgeon to change a lightbulb.
"What does os.path.join do?" → Gemini Flash ($0.000001 — literally free)
"Refactor the auth module" → Claude Sonnet ($0.003)
"Design the full system arch" → Claude Opus ($0.015)
| Task type | Without Router | With Router | Savings |
|---|---|---|---|
| Simple queries (60% of work) | Opus — $0.015 | Haiku/Gemini Flash — $0.0001 | 99% |
| Moderate tasks (30% of work) | Opus — $0.015 | Sonnet — $0.003 | 80% |
| Complex tasks (10% of work) | Opus — $0.015 | Opus — $0.015 | 0% |
| Blended monthly estimate | ~$50/mo | ~$8–15/mo | 70–85% |
💡 With Ollama: simple tasks route to a free local model — those 60% of queries cost $0.
Quick Start
Zero API keys required — if you have a Claude Code subscription, the router works out of the box. Simple tasks route to Claude Haiku (included), complex ones escalate to Sonnet/Opus. External providers (GPT-4o, Gemini, Perplexity) are optional add-ons.
One-line install
pipx install claude-code-llm-router && llm-router install
Or with pip:
pip install claude-code-llm-router && llm-router install
Option B: Claude Code Plugin
claude plugin add ypollak2/llm-router
Option C: Manual Install
git clone https://github.com/ypollak2/llm-router.git
cd llm-router
uv sync
Works with Cursor, Windsurf, and Zed
LLM Router is an MCP server — it works in any IDE that supports the Model Context Protocol.
Cursor — add to ~/.cursor/mcp.json:
{
"mcpServers": {
"llm-router": {
"command": "llm-router",
"args": []
}
}
}
Windsurf — add to ~/.codeium/windsurf/mcp_config.json:
{
"mcpServers": {
"llm-router": {
"command": "llm-router",
"args": []
}
}
}
Zed — add to Zed's settings.json:
{
"context_servers": {
"llm-router": {
"command": {
"path": "llm-router",
"args": []
}
}
}
}
The MCP tools (
llm_query,llm_code,llm_research, etc.) work identically in all IDEs. The auto-route hook is Claude Code-specific; other IDEs call the tools directly.
Enable Global Auto-Routing
Make the router evaluate every prompt across all projects:
# From the MCP tool:
llm_setup(action='install_hooks')
# Or from the CLI:
llm-router install
This installs hooks + rules to ~/.claude/ so every Claude Code session auto-routes tasks to the optimal model.
Start for free: Google's Gemini API has a free tier with 1M tokens/day. Groq also offers a generous free tier with ultra-fast inference.
What You Get
- 30 MCP tools — smart routing, text/code, image/video/audio, streaming, orchestration, usage monitoring, web dashboard
- Auto-route hook — intercepts every prompt before your top-tier model sees it; heuristic → Ollama → cheap API classifier chain, hooks self-update on
pip upgrade - Claude subscription mode — routes entirely within your CC subscription; Codex (free) before paid externals; external only when quota exhausted
- Anthropic prompt caching — auto-injects
cache_controlbreakpoints on long system prompts; up to 90% savings on repeated context - Semantic dedup cache — Ollama embeddings + cosine similarity skip identical-intent calls at zero cost
- Web dashboard —
llm-router dashboard→localhost:7337; cost trends, model distribution, recent decisions - Hard spend caps —
LLM_ROUTER_DAILY_SPEND_LIMITandLLM_ROUTER_MONTHLY_BUDGETraise before any call - Prompt classification cache — SHA-256 LRU cache for instant repeat classifications
- Circuit breaker + health — catches 429s, marks unhealthy providers, auto-recovers
- Quality logging — records every routing decision;
llm_quality_reportshows accuracy, savings, downshift rate - Cross-platform — macOS, Linux, Windows (desktop notifications, background processes, path handling)
Dashboard
The built-in web dashboard (llm_dashboard or llm-router dashboard) gives you a live view of routing decisions, cost trends, and subscription pressure.
| Overview | Performance |
|---|---|
| Logs & Analysis |
|---|
Design: Liquid Glass dark theme — Inter + JetBrains Mono, Material Symbols, Tailwind CSS. Auto-refreshes every 30 s.
How It Works
Auto-Route Hook — Every Prompt, Cheaper Model First
The UserPromptSubmit hook intercepts all prompts before your top-tier model sees them.
| Prompt | Classified as | Model used |
|---|---|---|
why doesn't the router work? |
analyze/moderate |
Haiku |
how does benchmarks.py work? |
query/simple |
Ollama / Haiku |
fix the bug in profiles.py |
code/moderate |
Haiku / Sonnet |
implement a distributed cache |
code/complex |
Sonnet / Opus |
write a blog post about LLMs |
generate/moderate |
Haiku / Gemini Flash |
git status (raw shell command) |
(skipped — terminal op) | — |
Classification chain (stops at first success):
1. Heuristic scoring instant, free → high-confidence patterns route immediately
2. Ollama local LLM free, ~1s → catches what heuristics miss
3. Cheap API ~$0.0001 → Gemini Flash / GPT-4o-mini fallback
4. Query catch-all instant, free → any remaining question → Haiku
Hook scripts are versioned and self-update — existing users get improvements automatically after pip install --upgrade.
Claude Code Subscription Mode
If you use Claude Code Pro/Max, you already pay for Haiku, Sonnet, and Opus. Enable subscription mode and the router routes within your subscription first — Codex (free via OpenAI subscription) before any paid API call, external only when quota is exhausted.
# In .env
LLM_ROUTER_CLAUDE_SUBSCRIPTION=true
Default Routing (No Pressure)
| Complexity | Model | Cost |
|---|---|---|
| simple | Claude Haiku 4.5 | free (subscription) |
| moderate | Sonnet (passthrough) | free (you're already using it) |
| complex | Claude Opus 4.6 | free (subscription) |
| research | Perplexity Sonar Pro | ~$0.005/query |
Pressure Cascade
| Condition | simple | moderate | complex |
|---|---|---|---|
| session < 95%, sonnet < 95% | Haiku (sub) | Sonnet (sub) | Opus (sub) |
| sonnet ≥ 95% | Codex → external | Codex → external | Opus (sub) |
| weekly ≥ 95% or session ≥ 95% | Codex → external | Codex → external | Codex → external |
Run llm_check_usage at session start to populate accurate pressure data. Hooks flag ⚠️ STALE when usage data is >30 minutes old.
External Fallback Chains (free-first)
| Tier | Chain |
|---|---|
| BUDGET (simple) | Ollama → Codex/gpt-5.4 → Codex/o3 → Gemini Flash → Groq → GPT-4o-mini |
| BALANCED (moderate) | Ollama → Codex/gpt-5.4 → Codex/o3 → GPT-4o → Gemini Pro → DeepSeek |
| PREMIUM (complex) | Ollama → Codex/gpt-5.4 → Codex/o3 → o3 → Gemini Pro |
Live subscription status:
+----------------------------------------------------------+
| Claude Subscription (Live) |
+----------------------------------------------------------+
| Session [====........] 35% resets in 3h 7m |
| Weekly (all) [===.........] 23% resets Fri 01:00 PM |
| Sonnet only [===.........] 26% resets Wed 10:00 AM |
+----------------------------------------------------------+
| OK 35% pressure -- full model selection |
+----------------------------------------------------------+
Providers
| Provider | Models | Free Tier | Best For |
|---|---|---|---|
| 🦙 Ollama | Any local model | Yes (free forever) | Privacy, zero cost, offline |
| Google Gemini | 2.5 Pro, 2.5 Flash | Yes (1M tokens/day) | Generation, long context |
| Groq | Llama 3.3, Mixtral | Yes | Ultra-fast inference |
| OpenAI | GPT-4o, GPT-4o-mini, o3 | No | Code, analysis, reasoning |
| Perplexity | Sonar, Sonar Pro | No | Research, current events |
| Anthropic | Claude Sonnet, Haiku | No | Nuanced writing, safety |
| Deepseek | V3, Reasoner | Yes (limited) | Cost-effective reasoning |
| Mistral | Large, Small | Yes (limited) | Multilingual |
| Together | Llama 3, CodeLlama | Yes (limited) | Open-source models |
| xAI | Grok 3 | No | Real-time information |
| Cohere | Command R+ | Yes (trial) | RAG, enterprise search |
Image, video, and audio providers (fal.ai, Runway, Stability AI, ElevenLabs, etc.) — see docs/PROVIDERS.md for full setup guides.
🦙 Ollama runs models locally — no API key, no cost, no data sent externally. Setup guide →
MCP Tools
Once installed, Claude Code gets these 29 tools:
| Tool | What It Does |
|---|---|
| Smart Routing | |
llm_classify |
Classify complexity + recommend model with time-aware budget pressure |
llm_route |
Auto-classify, then route to the best external LLM |
llm_track_usage |
Report Claude Code token usage for budget tracking |
llm_stream |
Stream LLM responses for long-running tasks |
| Text & Code | |
llm_query |
General questions — auto-routed to the best text LLM |
llm_research |
Search-augmented answers via Perplexity |
llm_generate |
Creative content — writing, summaries, brainstorming |
llm_analyze |
Deep reasoning — analysis, debugging, problem decomposition |
llm_code |
Coding tasks — generation, refactoring, algorithms |
llm_edit |
Route code-edit reasoning to a cheap model → returns exact {file, old_string, new_string} pairs |
| Media | |
llm_image |
Image generation — Gemini Imagen, DALL-E, Flux, or SD |
llm_video |
Video generation — Gemini Veo, Runway, Kling, etc. |
llm_audio |
Voice/audio — TTS via ElevenLabs or OpenAI |
| Orchestration | |
llm_orchestrate |
Multi-step pipelines across multiple models |
llm_pipeline_templates |
List available orchestration templates |
| Monitoring & Setup | |
llm_check_usage |
Check live Claude subscription usage (session %, weekly %) |
llm_update_usage |
Feed live usage data from claude.ai into the router |
llm_refresh_claude_usage |
Force-refresh Claude subscription data via OAuth |
llm_codex |
Route tasks to local Codex desktop agent (free, uses OpenAI sub) |
llm_setup |
Discover API keys, add providers, validate keys, install global hooks |
llm_rate |
Rate last response (👍/👎) — stored in routing_decisions for quality tracking |
llm_quality_report |
Routing accuracy, classifier stats, savings metrics, downshift rate |
llm_set_profile |
Switch routing profile (budget / balanced / premium) |
llm_usage |
Unified dashboard — Claude sub, Codex, APIs, savings in one view |
llm_health |
Check provider availability and circuit breaker status |
llm_providers |
List all supported and configured providers |
llm_cache_stats |
View cache hit rate, entries, memory estimate, evictions |
llm_cache_clear |
Clear the classification cache |
| Session Memory | |
llm_save_session |
Summarize + persist current session for cross-session context injection |
Context injection: text tools (
llm_query,llm_research,llm_generate,llm_analyze,llm_code) automatically prepend recent conversation history to every external call — GPT-4o, Gemini, and Perplexity receive the same context you have. Controlled byLLM_ROUTER_CONTEXT_ENABLED(default: on).
Routing Profiles
Three built-in profiles map to task complexity. Switch anytime:
llm_set_profile("budget") # Development, drafts, exploration
llm_set_profile("balanced") # Production work, client deliverables
llm_set_profile("premium") # Critical tasks, maximum quality
| Budget (simple) | Balanced (moderate) | Premium (complex) | |
|---|---|---|---|
| Text | Ollama → Haiku → Gemini Flash | Sonnet → GPT-4o → DeepSeek | Opus → Sonnet → o3 |
| Code | Ollama → Codex → DeepSeek → Haiku | Codex → Sonnet → GPT-4o | Codex → Opus → o3 |
| Research | Perplexity Sonar | Perplexity Sonar Pro | Perplexity Sonar Pro |
| Image | Flux Dev, Imagen Fast | Flux Pro, Imagen 3, DALL-E 3 | Imagen 3, DALL-E 3 |
| Video | minimax, Veo 2 | Kling, Veo 2, Runway Turbo | Veo 2, Runway Gen-3 |
| Audio | OpenAI TTS | ElevenLabs | ElevenLabs |
Model order is pressure-aware — as Claude quota is consumed, chains reorder to preserve remaining budget. See BENCHMARKS.md for how model quality scores drive rankings.
Budget Control
# In .env
LLM_ROUTER_MONTHLY_BUDGET=50 # USD, 0 = unlimited
The router tracks real-time spend across all providers in SQLite and blocks requests when the monthly budget is reached.
llm_usage("month")
→ Calls: 142 | Tokens: 320,000 | Cost: $3.42 | Budget: 6.8% of $50
Per-provider budgets: LLM_ROUTER_BUDGET_OPENAI=10.00, LLM_ROUTER_BUDGET_GEMINI=5.00.
Multi-Step Orchestration
Chain tasks across models in a pipeline:
llm_orchestrate("Research AI trends and write a report", template="research_report")
| Template | Pipeline |
|---|---|
research_report |
Research → Analyze → Write |
competitive_analysis |
Multi-source research → SWOT → Report |
content_pipeline |
Research → Draft → Review → Polish |
code_review_fix |
Review → Fix → Test |
Configuration
# Required: at least one provider
GEMINI_API_KEY=AIza... # Free tier! https://aistudio.google.com/apikey
OPENAI_API_KEY=sk-proj-...
PERPLEXITY_API_KEY=pplx-...
# Optional: more providers
ANTHROPIC_API_KEY=sk-ant-...
DEEPSEEK_API_KEY=...
GROQ_API_KEY=gsk_...
FAL_KEY=...
ELEVENLABS_API_KEY=...
# Router config
LLM_ROUTER_PROFILE=balanced # budget | balanced | premium
LLM_ROUTER_MONTHLY_BUDGET=0 # USD, 0 = unlimited
LLM_ROUTER_CLAUDE_SUBSCRIPTION=false # true = Claude Code Pro/Max user
# Ollama (two independent roles — classifier and task answerer)
LLM_ROUTER_OLLAMA_URL=http://localhost:11434 # hook classifier
LLM_ROUTER_OLLAMA_MODEL=qwen3.5:latest
OLLAMA_BASE_URL=http://localhost:11434 # router answerer
OLLAMA_BUDGET_MODELS=llama3.2,qwen2.5-coder:7b
# Smart routing (Claude Code model selection)
QUALITY_MODE=balanced # best | balanced | conserve
MIN_MODEL=haiku # floor: haiku | sonnet | opus
See .env.example for the full list.
Ollama note:
LLM_ROUTER_OLLAMA_URLis for the hook classifier (classifying complexity);OLLAMA_BASE_URLis for the router answerer (actually answering tasks). Configuring one does not enable the other. See docs/PROVIDERS.md for the full local-first setup.
Development
uv sync --extra dev
uv run pytest tests/ -q --ignore=tests/test_integration.py
uv run ruff check src/ tests/
llm-router install # deploy hooks to ~/.claude/
See CLAUDE.md for architecture, module layout, and contribution guidelines.
Roadmap
See CHANGELOG.md for what's been shipped. Coming next:
| Version | Theme | Headline features |
|---|---|---|
| ✅ Web dashboard, prompt caching, semantic dedup cache, hard daily cap, cross-platform notifications | ||
| v1.4 | Routing Intelligence | Task-aware model preferences, reasoning model tier, learned routing |
| v1.5 | Agentic & Team | Agent-tree budget tracking, multi-user profiles, YAML pipelines |
| v2.0 | Learning Router | Self-improving classifier trained on your own routing history |
See ROADMAP.md for design notes and competitive context.
Contributing
See CONTRIBUTING.md. Key areas: new provider integrations, routing intelligence, MCP client testing, documentation.
License
MIT — use it however you want.
Project details
Release history Release notifications | RSS feed
Download files
Download the file for your platform. If you're not sure which to choose, learn more about installing packages.
Source Distribution
Built Distribution
Filter files by name, interpreter, ABI, and platform.
If you're not sure about the file name format, learn more about wheel file names.
Copy a direct link to the current filters
File details
Details for the file claude_code_llm_router-1.4.1.tar.gz.
File metadata
- Download URL: claude_code_llm_router-1.4.1.tar.gz
- Upload date:
- Size: 345.6 kB
- Tags: Source
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.8.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
6eb6e0e38a33e8a1251191e07d94fdb32511e68b712dd9e72737a95d212014cb
|
|
| MD5 |
74f71d827d05268b1e9cae3b464df642
|
|
| BLAKE2b-256 |
aba2389af416cb87f523919e6d25682f15166eaf192b7d4c5389428162d730c6
|
File details
Details for the file claude_code_llm_router-1.4.1-py3-none-any.whl.
File metadata
- Download URL: claude_code_llm_router-1.4.1-py3-none-any.whl
- Upload date:
- Size: 212.7 kB
- Tags: Python 3
- Uploaded using Trusted Publishing? No
- Uploaded via: uv/0.8.3
File hashes
| Algorithm | Hash digest | |
|---|---|---|
| SHA256 |
519e5ee57b0a264f83861db6545a80cdc59779346ab9b95115f94edf5282b1ea
|
|
| MD5 |
136f29ab966baf9b18b36eb6b0507c0d
|
|
| BLAKE2b-256 |
12a0a97b6bc1bb7b132f77c0a258463f8e74a07f0f6aeed56b9f205bfa0aabb4
|